Sparsity and smoothness via the fused lasso
نویسندگان
چکیده
The lasso penalizes a least squares regression by the sum of the absolute values (L1-norm) of the coefficients. The form of this penalty encourages sparse solutions (with many coefficients equal to 0).We propose the ‘fused lasso’, a generalization that is designed for problems with features that can be ordered in some meaningful way. The fused lasso penalizes the L1-norm of both the coefficients and their successive differences. Thus it encourages sparsity of the coefficients and also sparsity of their differences—i.e. local constancy of the coefficient profile.The fused lasso is especially useful when the number of features p is much greater than N , the sample size.The technique is also extended to the ‘hinge’ loss function that underlies the support vector classifier.We illustrate the methods on examples from protein mass spectroscopy and gene expression data.
منابع مشابه
Predicting Workplace Incidents with Temporal Graph-guided Fused Lasso
We consider the problem of learning a sparse structured regression that utilizes both temporal information and structural information in the data. Our goal is to enforce sparsity in both the coefficients and their temporal differences, which is desirable for applications with data that evolves over time. We propose a regularization function based on fused lasso penalty that encourages temporal ...
متن کاملA Survey on Evolutionary Co-Clustering Formulations for Mining Time-Varying Data Using Sparsity Learning
The data matrix is considered as static in Traditional clustering and feature selection methods. However, the data matrices evolve smoothly over time in many applications. A simple approach to learn from these time-evolving data matrices is to analyze them separately. Such strategy ignores the time-dependent nature of the underlying data. Two formulations are proposed for evolutionary co-cluste...
متن کاملNon-parametric Group Orthogonal Matching Pursuit for Sparse Learning with Multiple Kernels
We consider regularized risk minimization in a large dictionary of Reproducing kernel Hilbert Spaces (RKHSs) over which the target function has a sparse representation. This setting, commonly referred to as Sparse Multiple Kernel Learning (MKL), may be viewed as the non-parametric extension of group sparsity in linear models. While the two dominant algorithmic strands of sparse learning, namely...
متن کاملProperties and Refinements of The Fused Lasso
We consider estimating an unknown signal, which is both blocky and sparse, corrupted by additive noise. We study three interrelated least squares procedures and their asymptotic properties. The first procedure is the fused lasso, put forward by Friedman et al. (2007), which we modify into a different estimator, called the fused adaptive lasso, with better properties. The other two estimators we...
متن کاملStructured Sparsity via Alternating Direction Methods ∗ Zhiwei ( Tony
We consider a class of sparse learning problems in high dimensional feature space regularized by a structured sparsity-inducing norm that incorporates prior knowledge of the group structure of the features. Such problems often pose a considerable challenge to optimization algorithms due to the non-smoothness and non-separability of the regularization term. In this paper, we focus on two commonl...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2004